Bandwidth Choice for Nonparametric Classification
نویسندگان
چکیده
It is shown that, for kernel-based classification with univariate distributions and two populations, optimal bandwidth choice has a dichotomous character. If the two densities cross at just one point, where their curvatures have the same signs, then minimum Bayes risk is achieved using bandwidths which are an order of magnitude larger than those which minimize pointwise estimation error. On the other hand, if the curvature signs are different, or if there are multiple crossing points, then bandwidths of conventional size are generally appropriate. The range of different modes of behavior is narrower in multivariate settings. There, the optimal size of bandwidth is generally the same as that which is appropriate for pointwise density estimation. These properties motivate empirical rules for bandwidth choice.
منابع مشابه
Variable data driven bandwidth choice in nonparametric quantile regression
The choice of a smoothing parameter or bandwidth is crucial when applying nonparametric regression estimators. In nonparametric mean regression various methods for bandwidth selection exists. But in nonparametric quantile regression bandwidth choice is still an unsolved problem. In this paper a selection procedure for local varying bandwidths based on the asymptotic mean squared error (MSE) of ...
متن کاملPenalizing function based bandwidth choice in nonparametric quantile regression
Abstract: In nonparametric mean regression various methods for bandwidth choice exist. These methods can roughly be divided into plug-in methods and methods based on penalizing functions. This paper uses the approach based on penalizing functions and adapt it to nonparametric quantile regression estimation, where bandwidth choice is still an unsolved problem. Various criteria for bandwitdth cho...
متن کاملEmpirical Bias Bandwidth Choice for Local Polynomial Matching Estimators
This paper addresses the choice of an optimal smoothing parameter for local polynomial matching. A version of Empirical Bias Bandwidth Selection (EBBS) proposed by Ruppert (1997) is applied to account for the MSE computation of the matching estimator. Thereby, an estimator for the large sample variance of the local polynomial matching estimator is also provided. A Monte Carlo study indicates be...
متن کاملA Comparative Study of Bandwidth Choice in Kernel Density Estimation for Naive Bayesian Classification
Kernel density estimation (KDE) is an important method in nonparametric learning. While KDE has been studied extensively in the context of accuracy of density estimation, it has not been studied extensively in the context of classification. This paper studies nine bandwidth selection schemes for kernel density estimation in Naive Bayesian classification context, using 52 machine learning benchm...
متن کاملOn adaptive smoothing in kernel discriminant analysis
One popular application of kernel density estimation is in kernel discriminant analysis, where kernel estimates of population densities are plugged in the Bayes rule to develop a nonparametric classifier. Performance of these kernel density estimates and that of the corresponding classifier depend on the values of associated smoothing parameters commonly known as the bandwidths. Bandwidths that...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2005